Goto

Collaborating Authors

 unconstrained generation order



Reviews: Sequence Modeling with Unconstrained Generation Order

Neural Information Processing Systems

Updated review: The authors have indicated that they will run additional experiments and make the clarifications I requested, so I will raise my score in 7 in agreement with the other reviews leading to an "accept" consensus. However, I do note that in their rebuttal the authors describe Gu et al., Stern et al., and Welleck et al. as "concurrent work". To be totally clear, all three of those papers were posted to arxiv in early February; the NeurIPS deadline was over 3 months later and it is now 6 months after they papers appeared online. I would argue that 3 (or 6) months is long enough to provide a more direct comparison and would not consider this submission "concurrent work". I don't think this warrants rejecting the paper, but I do want to note that I disagree with the authors here and still believe that a more direct comparison is appropriate.


Sequence Modeling with Unconstrained Generation Order

Neural Information Processing Systems

The dominant approach to sequence generation is to produce a sequence in some predefined order, e.g. In contrast, we propose a more general model that can generate the output sequence by inserting tokens in any arbitrary order. Our model learns decoding order as a result of its training procedure. Our experiments show that this model is superior to fixed order models on a number of sequence generation tasks, such as Machine Translation, Image-to-LaTeX and Image Captioning.


Sequence Modeling with Unconstrained Generation Order

Emelianenko, Dmitrii, Voita, Elena, Serdyukov, Pavel

Neural Information Processing Systems

The dominant approach to sequence generation is to produce a sequence in some predefined order, e.g. In contrast, we propose a more general model that can generate the output sequence by inserting tokens in any arbitrary order. Our model learns decoding order as a result of its training procedure. Our experiments show that this model is superior to fixed order models on a number of sequence generation tasks, such as Machine Translation, Image-to-LaTeX and Image Captioning. Papers published at the Neural Information Processing Systems Conference.